Goto

Collaborating Authors

 graph representation


Permutation-InvariantVariationalAutoencoderfor Graph-LevelRepresentationLearning

Neural Information Processing Systems

Most work, however, focuses on either node-or graph-level supervised learning, such as node, link or graph classification or node-level unsupervised learning (e.g., node clustering). Despite its wide range of possible applications, graph-level unsupervised representation learning has not received much attention yet. This might be mainly attributed to the high representation complexity ofgraphs, which can berepresented byn!equivalent adjacencymatrices, where n is the number of nodes. In this work we address this issue by proposing a permutation-invariant variational autoencoder for graph structured data.









7 Appendix Figure 5: Comparison of GenStat architecture to selected graph generative models. 7.1 Proofs 7.1.1 Proposition 1 Let p

Neural Information Processing Systems

Figure 5: Comparison of GenStat architecture to selected graph generative models. This proof uses two properties of LDP: composability and immunity to post-processing [2]. Figure 6 illustrates the PGM of Randomized algorithms. The GGM parameters are a function of the perturbed graph statistics as learning input. The implementation can be easily extended to directed graphs. A statistics-based GGM that takes the degree sequence as sufficient statistics [5].